Shai Ben - David – Research Plan

نویسنده

  • Shai Ben-David
چکیده

The complexity of machine learning tasks can be measured along two parameters; Information and computation. Information complexity is concerned with the generalization performance of learning algorithms. Namely, How many training examples are needed to achieve certain accuracy of prediction? How fast do empirical estimates converge to the true population parameters? Etc. The computational complexity of learning concerns the computation applied to the training data in order to extract from it the learners predictions. Examining the wide variety of analytic results in the area, there seems to be a tradeoff between these two complexity parameters. Learning paradigms that enjoy computationally efficient algorithms fail to have satisfactory generalization guarantees, and vice versa. Although this phenomena is manifested across the board of machine learning theory, we are far from understanding it, let alone from being able to prove any rigorous tradeoff results. I view this question as one of the most basic and intriguing challenges in the area of computational statistics. I have been working on related issues in the past, and I plan to carry out related research in the future. Some of the topics I list below can be naturally viewed as different aspects of the above basic issue.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

How Far Are We From Having a Satisfactory Theory of Clustering?

This is an overview of the invited talk delivered at the 41st International Symposium on Mathematical Foundations of Computer Science (MFCS-2016). 1998 ACM Subject Classification I.5.3 Clustering

متن کامل

Weighted Clustering

In this paper we investigate clustering in the weighted setting, in which every data point is assigned a real valued weight. We conduct a theoretical analysis on the influence of weighted data on standard clustering algorithms in each of the partitional and hierarchical settings, characterising the precise conditions under which such algorithms react to weights, and classifying clustering metho...

متن کامل

Minimizing The Misclassification Error Rate Using a Surrogate Convex Loss

We carefully study how well minimizing convex surrogate loss functions corresponds to minimizing the misclassification error rate for the problem of binary classification with linear predictors. We consider the agnostic setting, and investigate guarantees on the misclassification error of the loss-minimizer in terms of the margin error rate of the best predictor. We show that, aiming for such a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007